Projection Pursuit via Decomposition of Bias Termsof Kernel Density
نویسندگان
چکیده
Dimension reduction of data, < d ! < p (p << d), to be used for clustering has speciic requirements that are not generally met by generic dimension reduction algorithms such as principal components. Projection pursuit, on the other hand, has a growing variety of criteria that target holes, skewness, etc., using information measures, density functionals, sample moments, etc. With the exception of q = 3 by Nason (19xx), the target dimension is only one or two. In this paper, we focus on features of a multivariate kernel density estimator to nd a sequence of basis vectors deening our subspace. We begin by deriving two new variations on the formula for obtaining the asymptotic integrated squared bias (AISB) of a general multivariate kernel density estimator. These formulae are then used to compute orthogonal projections of data onto lower dimensional subspaces in which as much as possible of the AISB is retained. The methods we use for dimension reduction favor those subspaces on which the clustering information of the data is prominent. The methods are not of the 1 \least normal" variety, since the normal density is not the \easiest" estimable density, but rather a certain low order polynomial density.
منابع مشابه
The Relative Improvement of Bias Reduction in Density Estimator Using Geometric Extrapolated Kernel
One of a nonparametric procedures used to estimate densities is kernel method. In this paper, in order to reduce bias of kernel density estimation, methods such as usual kernel(UK), geometric extrapolation usual kernel(GEUK), a bias reduction kernel(BRK) and a geometric extrapolation bias reduction kernel(GEBRK) are introduced. Theoretical properties, including the selection of smoothness para...
متن کاملComparison of the Gamma kernel and the orthogonal series methods of density estimation
The standard kernel density estimator suffers from a boundary bias issue for probability density function of distributions on the positive real line. The Gamma kernel estimators and orthogonal series estimators are two alternatives which are free of boundary bias. In this paper, a simulation study is conducted to compare small-sample performance of the Gamma kernel estimators and the orthog...
متن کاملThe Partially Linear Regression Model: Monte Carlo Evidence from the Projection Pursuit Regression Approach
In a partially linear regression model with a high dimensional unknown component we find an estimator of the parameter of the linear part based on projection pursuit methods to be considerably more efficient than the standard density weighted kernel estimator.
متن کاملNonparametric multivariate density estimation: a comparative study
This paper algorithmically and empirically studies two major types of nonparametric multivariate density estimation techniques, where no assumption is made about the data being drawn from any of known parametric families of distribution. The first type is the popular kernel method (and several of its variants) which uses locally tuned radial basis (e.g., Gaussian) functions to interpolate the m...
متن کاملFast optimal bandwidth selection for kernel density estimation
We propose a computationally efficient 2−exact approximation algorithm for univariate Gaussian kernel based density derivative estimation that reduces the computational complexity from O(MN) to linear O(N +M). We apply the procedure to estimate the optimal bandwidth for kernel density estimation. We demonstrate the speedup achieved on this problem using the ”solve-the-equation plug-in” method, ...
متن کامل